Estimation of Covariance Matrices under Sparsity Constraints

نویسندگان

  • Philippe Rigollet
  • Alexandre B. Tsybakov
چکیده

Discussion of “Minimax Estimation of Large Covariance Matrices under L1-Norm” by Tony Cai and Harrison Zhou. To appear in Statistica Sinica. Introduction. Estimation of covariance matrices in various norms is a critical issue that finds applications in a wide range of statistical problems, and especially in principal component analysis. It is well known that, without further assumptions, the empirical covariance matrix Σ is the best possible estimator in many ways, and in particular in a minimax sense. However, it is also well known that Σ is not an accurate estimator when the dimension p of the observations is high. The minimax analysis carried out by Tony Cai and Harry Zhou ([CZ] in what follows) guarantees that for several classes of matrices with reasonable structure (sparse or banded matrices), the fully data-driven thresholding estimator achieves the best possible rates when p is much larger than the sample size n. This is done, in particular, by proving minimax lower bounds that ensure that no estimator can perform better than the hard thresholding estimator, uniformly over the sparsity classes Gq for each 0 ≤ q < 1. This result has a flavor of universality in the sense that one and the same estimator is minimax optimal for several classes of matrices. Our comments focus on the sparsity classes of matrices. (a) Optimal rates. Optimal rates are obtained in [CZ] under the assumption that the dimension is very high: p ≥ nν, ν > 1. Thus, the case of dimensions smaller than n, or even p ≈ n, is excluded. This seems to be due to the technique of proving the lower bound (Theorem 2 in [CZ]). Indeed, by a different technique, we show that the lower bound holds without this assumption, cf. Theorem 1 below. Furthermore, in general, our lower rate ψ(1) is different from that obtained in [CZ] and has ingredients similar to the optimal rate for the Gaussian sequence model. We conjecture that it is optimal for all admissible configurations of n, p, and sparsity parameters. (b) Frobenius norm and global sparsity. We argue that the Frobenius norm is naturally adapted to the structure of the problem, at least for Gaussian observations, and we derive optimal rates under the Frobenius risk and global sparsity assumption. Department of Operations Research and Financial Engineering, Princeton University, Princeton, NJ 08544, USA. Supported in part by NSF grants DMS-0906424 and CAREER-DMS-1053987. Laboratoire de Statistique, CREST-ENSAE, 3, av. Pierre Larousse, F-92240 Malakoff Cedex, France. Supported in part by ANR under the grant “Parcimonie”.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Sparse estimation of high-dimensional correlation matrices

Estimating covariations of variables for high dimensional data is important for understanding their relations. Recent years have seen several attempts to estimate covariance matrices with sparsity constraints. A new convex optimization formulation for estimating correlation matrices, which are scale invariant, is proposed as opposed to covariance matrices. The constrained optimization problem i...

متن کامل

Minimax Estimation of Bandable Precision Matrices

The inverse covariance matrix provides considerable insight for understanding statistical models in the multivariate setting. In particular, when the distribution over variables is assumed to be multivariate normal, the sparsity pattern in the inverse covariance matrix, commonly referred to as the precision matrix, corresponds to the adjacency matrix representation of the Gauss-Markov graph, wh...

متن کامل

PAGE: Robust Pattern Guided Estimation of Large Covariance Matrix

We study the problem of estimating large covariance matrices under two types of structural assumptions: (i) The covariance matrix is the summation of a low rank matrix and a sparse matrix, and we have some prior information on the sparsity pattern of the sparse matrix; (ii) The data follow a transelliptical distribution. The former structure regulates the parameter space and has its roots in di...

متن کامل

Covariance Matrix Estimation for Stationary Time Series

We obtain a sharp convergence rate for banded covariance matrix estimates of stationary processes. A precise order of magnitude is derived for spectral radius of sample covariance matrices. We also consider a thresholded covariance matrix estimator that can better characterize sparsity if the true covariance matrix is sparse. As our main tool, we implement Toeplitz [Math. Ann. 70 (1911) 351–376...

متن کامل

0 Sparse Inverse Covariance Estimation

Recently, there has been focus on penalized loglikelihood covariance estimation for sparse inverse covariance (precision) matrices. The penalty is responsible for inducing sparsity, and a very common choice is the convex l1 norm. However, the best estimator performance is not always achieved with this penalty. The most natural sparsity promoting “norm” is the non-convex l0 penalty but its lack ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2012